A p value from the Hosmer-Lemeshow (H-L) test: In Figure 18-4a, this is listed under Hosmer-
Lemeshow Goodness of Fit Test. The null hypothesis for this test is your data are consistent with
the logistic function’s S shape, so if p < 0.05, your data do not qualify for logistic regression. The
focus of the test is to see if the S is getting distorted at very high or very low levels of the predictor
(as shown in Figure 18-4b). In Figure 18-4a, the H-L p value is 0.842, which means that the data
are consistent with the shape of a logistic curve.
One or more pseudo–r2values:Pseudo–r2values indicate how much of the total variability in the
outcome is explainable by the fitted model. They are analogous to how r2 is interpreted in ordinary
least-squares regression, as described in Chapter 17. In Figure 18-4a, two such values are
provided under the labels Cox/Snell R-square and Nagelkerke R-square. The Cox/Snell r2 is
0.577, and the Nagelkerke r2 is 0.770, both of which indicate that a majority of the variability in
the outcome is explainable by the logistic model.
Akaike’s Information Criterion (AIC): AIC is a measure of the final model deviance
adjusted for how many predictor variables are in the model. Like deviance, the smaller the AIC,
the better the fit. The AIC is not very useful on its own, and is instead used for choosing between
different models. When all the predictors in one model are nested — or included — in another
model with more predictors, the AIC is helpful for comparing these models to see if it is worth
adding the extra predictors.
Checking out the table of regression coefficients
Your intention when developing a logistic regression model is to obtain estimates from the
table of coefficients, which looks much like the coefficients table from ordinary straight-line or
multivariate least-squares regression (see Chapters 16 and 17). In Figure 18-4a, they are listed
under Coefficients and Standard Errors. Observe:
Every predictor variable appears on a separate row.
There’s one row for the constant term labeled Intercept.
The first column usually lists the regression coefficients (under Coeff. in Figure 18-4a).
The second column usually lists the standard error (SE) of each coefficient (under StdErr in Figure
18-4a).
A p-value column indicates whether the coefficient is statistically significantly different from 0.
This column may be labeled Sig or Signif or Pr(> lzl), but in Figure 18-4a, it is labeled p-value.
For each predictor variable, the output should also provide the odds ratio (OR) and its 95 percent
confidence interval. These are usually presented in a separate table as they are in Figure 18-4a under
Odds Ratios and 95% Confidence Intervals.
Predicting probabilities with the fitted logistic formula